Goto

Collaborating Authors

 procedural model


Neurally-Guided Procedural Models: Amortized Inference for Procedural Graphics Programs using Neural Networks

Neural Information Processing Systems

Probabilistic inference algorithms such as Sequential Monte Carlo (SMC) provide powerful tools for constraining procedural models in computer graphics, but they require many samples to produce desirable results. In this paper, we show how to create procedural models which learn how to satisfy constraints. We augment procedural models with neural networks which control how the model makes random choices based on the output it has generated thus far. We call such models neurally-guided procedural models. As a pre-computation, we train these models to maximize the likelihood of example outputs generated via SMC. They are then used as efficient SMC importance samplers, generating high-quality results with very few samples. We evaluate our method on L-system-like models with image-based constraints. Given a desired quality threshold, neurally-guided models can generate satisfactory results up to 10x faster than unguided models.



Separating Knowledge and Perception with Procedural Data

Rodríguez-Muñoz, Adrián, Baradad, Manel, Isola, Phillip, Torralba, Antonio

arXiv.org Artificial Intelligence

We train representation models with procedural data only, and apply them on visual similarity, classification, and semantic segmentation tasks without further training by using visual memory -- an explicit database of reference image embeddings. Unlike prior work on visual memory, our approach achieves full compartmentalization with respect to all real-world images while retaining strong performance. Compared to a model trained on Places, our procedural model performs within $1\%$ on NIGHTS visual similarity, outperforms by $8\%$ and $15\%$ on CUB200 and Flowers102 fine-grained classification, and is within $10\%$ on ImageNet-1K classification. It also demonstrates strong zero-shot segmentation, achieving an $R^2$ on COCO within $10\%$ of the models trained on real data. Finally, we analyze procedural versus real data models, showing that parts of the same object have dissimilar representations in procedural models, resulting in incorrect searches in memory and explaining the remaining performance gap.


Synthesizing 3D Abstractions by Inverting Procedural Buildings with Transformers

Dax, Maximilian, Berbel, Jordi, Stria, Jan, Guibas, Leonidas, Bergmann, Urs

arXiv.org Artificial Intelligence

Training is fully supervised, We generate abstractions of buildings, reflecting the essential based on a dataset of procedural buildings paired aspects of their geometry and structure, by learning with corresponding point cloud simulations. We develop to invert procedural models. We first build a dataset of various technical components tailored to the generation of abstract procedural building models paired with simulated abstractions. This includes the design of a programmatic point clouds and then learn the inverse mapping through a language to efficiently represent abstractions, its combination transformer. Given a point cloud, the trained transformer with a technique to guarantee transformer outputs consistent then infers the corresponding abstracted building in terms with the structure imposed by this language, and an of a programmatic language description.


Generating Diverse Agricultural Data for Vision-Based Farming Applications

Cieslak, Mikolaj, Govindarajan, Umabharathi, Garcia, Alejandro, Chandrashekar, Anuradha, Hädrich, Torsten, Mendoza-Drosik, Aleksander, Michels, Dominik L., Pirk, Sören, Fu, Chia-Chun, Pałubicki, Wojciech

arXiv.org Artificial Intelligence

We present a specialized procedural model for generating synthetic agricultural scenes, focusing on soybean crops, along with various weeds. This model is capable of simulating distinct growth stages of these plants, diverse soil conditions, and randomized field arrangements under varying lighting conditions. The integration of real-world textures and environmental factors into the procedural generation process enhances the photorealism and applicability of the synthetic data. Our dataset includes 12,000 images with semantic labels, offering a comprehensive resource for computer vision tasks in precision agriculture, such as semantic segmentation for autonomous weed control. We validate our model's effectiveness by comparing the synthetic data against real agricultural images, demonstrating its potential to significantly augment training data for machine learning models in agriculture. This approach not only provides a cost-effective solution for generating high-quality, diverse data but also addresses specific needs in agricultural vision tasks that are not fully covered by general-purpose models.


Neurally-Guided Procedural Models: Amortized Inference for Procedural Graphics Programs using Neural Networks

Neural Information Processing Systems

Probabilistic inference algorithms such as Sequential Monte Carlo (SMC) provide powerful tools for constraining procedural models in computer graphics, but they require many samples to produce desirable results. In this paper, we show how to create procedural models which learn how to satisfy constraints. We augment procedural models with neural networks which control how the model makes random choices based on the output it has generated thus far. We call such models neurally-guided procedural models. As a pre-computation, we train these models to maximize the likelihood of example outputs generated via SMC. They are then used as efficient SMC importance samplers, generating high-quality results with very few samples. We evaluate our method on L-system-like models with imagebased constraints. Given a desired quality threshold, neurally-guided models can generate satisfactory results up to 10x faster than unguided models.


DeepTree: Modeling Trees with Situated Latents

Zhou, Xiaochen, Li, Bosheng, Benes, Bedrich, Fei, Songlin, Pirk, Sören

arXiv.org Artificial Intelligence

In this paper, we propose DeepTree, a novel method for modeling trees based on learning developmental rules for branching structures instead of manually defining them. We call our deep neural model situated latent because its behavior is determined by the intrinsic state -- encoded as a latent space of a deep neural model -- and by the extrinsic (environmental) data that is situated as the location in the 3D space and on the tree structure. We use a neural network pipeline to train a situated latent space that allows us to locally predict branch growth only based on a single node in the branch graph of a tree model. We use this representation to progressively develop new branch nodes, thereby mimicking the growth process of trees. Starting from a root node, a tree is generated by iteratively querying the neural network on the newly added nodes resulting in the branching structure of the whole tree. Our method enables generating a wide variety of tree shapes without the need to define intricate parameters that control their growth and behavior. Furthermore, we show that the situated latents can also be used to encode the environmental response of tree models, e.g., when trees grow next to obstacles. We validate the effectiveness of our method by measuring the similarity of our tree models and by procedurally generated ones based on a number of established metrics for tree form.


Neural Synthesis of Footsteps Sound Effects with Generative Adversarial Networks

Comunità, Marco, Phan, Huy, Reiss, Joshua D.

arXiv.org Artificial Intelligence

To this day, there has not yet been an attempt at exploring the use of neural networks for the synthesis of footsteps sounds although Footsteps are among the most ubiquitous sound effects in multimedia there is substantial literature exploring neural synthesis of broadband applications. There is substantial research into understanding impulsive sounds, such as drums samples, which have some the acoustic features and developing synthesis models for footstep similarities to footsteps. One of the first attempts was in [15], where sound effects. In this paper, we present a first attempt at adopting Donahue et al. developed WaveGAN - a generative adversarial network neural synthesis for this task. We implemented two GAN-based architectures for unconditional audio synthesis. Another example of neural and compared the results with real recordings as well as synthesis of drums is [16], where the authors used a Progressive six traditional sound synthesis methods. Our architectures reached Growing GAN. Variational autoencoders [17] and U-Nets [18] realism scores as high as recorded samples, showing encouraging have also been used for the same task.


Neurally-Guided Procedural Models: Amortized Inference for Procedural Graphics Programs using Neural Networks

Ritchie, Daniel, Thomas, Anna, Hanrahan, Pat, Goodman, Noah

Neural Information Processing Systems

Probabilistic inference algorithms such as Sequential Monte Carlo (SMC) provide powerful tools for constraining procedural models in computer graphics, but they require many samples to produce desirable results. In this paper, we show how to create procedural models which learn how to satisfy constraints. We augment procedural models with neural networks which control how the model makes random choices based on the output it has generated thus far. We call such models neurally-guided procedural models. As a pre-computation, we train these models to maximize the likelihood of example outputs generated via SMC. They are then used as efficient SMC importance samplers, generating high-quality results with very few samples. We evaluate our method on L-system-like models with image-based constraints. Given a desired quality threshold, neurally-guided models can generate satisfactory results up to 10x faster than unguided models.


Neurally-Guided Procedural Models: Amortized Inference for Procedural Graphics Programs using Neural Networks

Ritchie, Daniel, Thomas, Anna, Hanrahan, Pat, Goodman, Noah D.

arXiv.org Artificial Intelligence

Probabilistic inference algorithms such as Sequential Monte Carlo (SMC) provide powerful tools for constraining procedural models in computer graphics, but they require many samples to produce desirable results. In this paper, we show how to create procedural models which learn how to satisfy constraints. We augment procedural models with neural networks which control how the model makes random choices based on the output it has generated thus far. We call such models neurally-guided procedural models. As a pre-computation, we train these models to maximize the likelihood of example outputs generated via SMC. They are then used as efficient SMC importance samplers, generating high-quality results with very few samples. We evaluate our method on L-system-like models with image-based constraints. Given a desired quality threshold, neurally-guided models can generate satisfactory results up to 10x faster than unguided models.